Supplementary Materials for “Naive Parallelization of Coordinate Descent Methods and an Application on Multi-core L1-regularized Classification”

نویسندگان

  • Yong Zhuang
  • Yuchin Juan
  • Guo-Xun Yuan
  • Chih-Jen Lin
چکیده

The subproblem is in a form of L1-regularized least squares. To perform a coordinate update in the CD subroutine, the training data is accessed/used in a feature-wise manner, which is the same as how data is used in CDN for problem (3.2). Thus, the applicability of the naive parallelization to the CD subroutine should hold here and the speedup is predicted to be at a comparable level. In addition, there are two other places to which we can apply the direct parallelization in newGLMNET. First, in preparation for solving (1.1) with CD, newGLMNET computes ∇L(w) and the diagonal entries of H beforehand. These operations can be parallelized directly. Second, newGLMNET requires a backtrack line search along a Newton step d where the function value decrease f(w + βd)− f(w) is computed for a given t. It is straightforward to parallelize each function evaluation through the naive approach. For more details of the newGLMNET algorithm, please see [4].

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Naive Parallelization of Coordinate Descent Methods and an Application on Multi-core L1-regularized Classification

It is well known that a direct parallelization of sequential optimization methods (e.g., coordinate descent and stochastic gradient methods) is often not effective. The reason is that at each iteration, the number of operations may be too small. In this paper, we point out that because of the skewed distribution of non-zero values in real-world data sets, this common understanding may not be tr...

متن کامل

Supplementary materials for "Parallel Dual Coordinate Descent Method for Large-scale Linear Classification in Multi-core Environments"

f(α) ≡ g(Eα) + bα, f(·) and g(·) are proper closed functions, E is a constant matrix, and Li ∈ [−∞,∞), Ui ∈ (−∞,∞] are lower/upper bounds. It has been checked in [1] that l1 and l2 loss SVM are in the form of (I.1) and satisfy additional assumptions needed in [4]. We introduce an important class of gradient-based scheme for CD’s variable selection: the Gauss-Southwell rule. It plays an importan...

متن کامل

A Comparison of Optimization Methods for Large-scale L1-regularized Linear Classification

Large-scale linear classification is widely used in many areas. The L1-regularized form can be applied for feature selection, but its non-differentiability causes more difficulties in training. Various optimization methods have been proposed in recent years, but no serious comparison among them has been made. In this paper, we discuss several state of the art methods and propose two new impleme...

متن کامل

A Comparison of Optimization Methods and Software for Large-scale L1-regularized Linear Classification

Large-scale linear classification is widely used in many areas. The L1-regularized form can be applied for feature selection; however, its non-differentiability causes more difficulties in training. Although various optimization methods have been proposed in recent years, these have not yet been compared suitably. In this paper, we first broadly review existing methods. Then, we discuss state-o...

متن کامل

Efficient parallelization of the genetic algorithm solution of traveling salesman problem on multi-core and many-core systems

Efficient parallelization of genetic algorithms (GAs) on state-of-the-art multi-threading or many-threading platforms is a challenge due to the difficulty of schedulation of hardware resources regarding the concurrency of threads. In this paper, for resolving the problem, a novel method is proposed, which parallelizes the GA by designing three concurrent kernels, each of which running some depe...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2017